Goto

Collaborating Authors

 Changzhou


SoK: Trust-Authorization Mismatch in LLM Agent Interactions

Shi, Guanquan, Du, Haohua, Wang, Zhiqiang, Liang, Xiaoyu, Liu, Weiwenpei, Bian, Song, Guan, Zhenyu

arXiv.org Artificial Intelligence

Large Language Models (LLMs) are rapidly evolving into autonomous agents capable of interacting with the external world, significantly expanding their capabilities through standardized interaction protocols. However, this paradigm revives the classic cybersecurity challenges of agency and authorization in a novel and volatile context. As decision-making shifts from deterministic code logic to probabilistic inference driven by natural language, traditional security mechanisms designed for deterministic behavior fail. It is fundamentally challenging to establish trust for unpredictable AI agents and to enforce the Principle of Least Privilege (PoLP) when instructions are ambiguous. Despite the escalating threat landscape, the academic community's understanding of this emerging domain remains fragmented, lacking a systematic framework to analyze its root causes. This paper provides a unifying formal lens for agent-interaction security. We observed that most security threats in this domain stem from a fundamental mismatch between trust evaluation and authorization policies. We introduce a novel risk analysis model centered on this trust-authorization gap. Using this model as a unifying lens, we survey and classify the implementation paths of existing, often seemingly isolated, attacks and defenses. This new framework not only unifies the field but also allows us to identify critical research gaps. Finally, we leverage our analysis to suggest a systematic research direction toward building robust, trusted agents and dynamic authorization mechanisms.


A Novel Deep Neural Network Architecture for Real-Time Water Demand Forecasting

Salloom, Tony, Kaynak, Okyay, He, Wei

arXiv.org Artificial Intelligence

Short-term water demand forecasting (StWDF) is the foundation stone in the derivation of an optimal plan for controlling water supply systems. Deep learning (DL) approaches provide the most accurate solutions for this purpose. However, they suffer from complexity problem due to the massive number of parameters, in addition to the high forecasting error at the extreme points. In this work, an effective method to alleviate the error at these points is proposed. It is based on extending the data by inserting virtual data within the actual data to relieve the nonlinearity around them. To our knowledge, this is the first work that considers the problem related to the extreme points. Moreover, the water demand forecasting model proposed in this work is a novel DL model with relatively low complexity. The basic model uses the gated recurrent unit (GRU) to handle the sequential relationship in the historical demand data, while an unsupervised classification method, k -means, is introduced for the creation of new features to enhance the prediction accuracy with less number of parameters. Real data obtained from two different water plants in China are used to train and verify the model proposed. The prediction results and the comparison with the state-of-the-art illustrate that the method proposed reduces the complexity of the model six times of what achieved in the literature while conserving the same accuracy. Furthermore, it is found that extending the data set significantly reduces the error by about 30%. However, it increases the training time. Introduction Water scarcity has become a threat to humankind in recent decades. Many efforts in all possible directions are being made to compensate for this growing problem (Northey et al., 2016; González-Zeas et al., 2019). The major reliable strategies for that include water treatment (Zinatloo-Ajabshir et al., 2020a), water desalination, and optimization of water management systems. Nanotechnology is the most powerful technology employed for water treatment, where researchers have done impressive work (Zinatloo-Ajabshir et al., 2020b, 2017; Moshtaghi et al., 2016). On the other hand, StWDF is the foundation stone of the optimization of water management systems.


Whatever Remains Must Be True: Filtering Drives Reasoning in LLMs, Shaping Diversity

Kruszewski, Germán, Erbacher, Pierre, Rozen, Jos, Dymetman, Marc

arXiv.org Artificial Intelligence

Reinforcement Learning (RL) has become the de facto standard for tuning LLMs to solve tasks involving reasoning. However, growing evidence shows that models trained in such way often suffer from a significant loss in diversity. We argue that this arises because RL implicitly optimizes the "mode-seeking" or "zero-forcing" Reverse KL to a target distribution causing the model to concentrate mass on certain high-probability regions of the target while neglecting others. In this work, we instead begin from an explicit target distribution, obtained by filtering out incorrect answers while preserving the relative probabilities of correct ones. Starting from a pre-trained LLM, we approximate this target distribution using the $α$-divergence family, which unifies prior approaches and enables direct control of the precision-diversity trade-off by interpolating between mode-seeking and mass-covering divergences. On a Lean theorem-proving benchmark, our method achieves state-of-the-art performance along the coverage-precision Pareto frontier, outperforming all prior methods on the coverage axis.


AudAgent: Automated Auditing of Privacy Policy Compliance in AI Agents

Zheng, Ye, Hu, Yidan

arXiv.org Artificial Intelligence

AI agents can autonomously perform tasks and, often without explicit user consent, collect or disclose users' sensitive local data, which raises serious privacy concerns. Although AI agents' privacy policies describe their intended data practices, there remains limited transparency and accountability about whether runtime behavior matches those policies. To close this gap, we introduce AudAgent, a visual tool that continuously monitors AI agents' data practices in real time and guards compliance with stated privacy policies. AudAgent consists of four components for automated privacy auditing of AI agents. (i) Policy formalization: a novel cross-LLM voting mechanism to guarantee confidence of the parsed privacy policy model. (ii) Runtime annotation: a lightweight Presidio-based analyzer detects sensitive data and annotates data practices based on the AI agent's context and the privacy policy model. (iii) Compliance auditing: ontology graphs and automata-based checking connect the privacy policy model with runtime annotations, enabling on-the-fly compliance checking. (iv) User interface: an infrastructure-independent implementation visualizes the real-time execution trace of AI agents along with potential privacy policy violations, providing user-friendly transparency and accountability. We evaluate AudAgent with AI agents built using mainstream frameworks, demonstrating its effectiveness in detecting and visualizing privacy policy violations in real time. Using AudAgent, we also find that most privacy policies omit explicit safeguards for highly sensitive data such as SSNs, whose misuse violates legal requirements, and that many agents do not refuse handling such data via third-party tools, including those controlled by Claude, Gemini, and DeepSeek. AudAgent proactively blocks operations on such data, overriding the agents' original privacy policy and behavior.


Cross-Modal Reconstruction Pretraining for Ramp Flow Prediction at Highway Interchanges

Li, Yongchao, Chen, Jun, Li, Zhuoxuan, Gao, Chao, Li, Yang, Zhang, Chu, Dong, Changyin

arXiv.org Artificial Intelligence

Interchanges are crucial nodes for vehicle transfers between highways, yet the lack of real-time ramp detectors creates blind spots in traffic prediction. To address this, we propose a Spatio-Temporal Decoupled Autoencoder (STDAE), a two-stage framework that leverages cross-modal reconstruction pretraining. In the first stage, STDAE reconstructs historical ramp flows from mainline data, forcing the model to capture intrinsic spatio-temporal relations. Its decoupled architecture with parallel spatial and temporal autoencoders efficiently extracts heterogeneous features. In the prediction stage, the learned representations are integrated with models such as GWNet to enhance accuracy. Experiments on three real-world interchange datasets show that STDAE-GWNET consistently outperforms thirteen state-of-the-art baselines and achieves performance comparable to models using historical ramp data. This demonstrates its effectiveness in overcoming detector scarcity and its plug-and-play potential for diverse forecasting pipelines.


Find Them All: Unveiling MLLMs for Versatile Person Re-identification

Li, Jinhao, Chen, Zijian, Deng, Lirong, Zhai, Guangtao, Wang, Changbo

arXiv.org Artificial Intelligence

Person re-identification (ReID) aims to retrieve images of a target person from the gallery set, with wide applications in medical rehabilitation and public security. However, traditional person ReID models are typically uni-modal, resulting in limited generalizability across heterogeneous data modalities. Recently, the emergence of multi-modal large language models (MLLMs) has shown a promising avenue for addressing this issue. Despite this potential, existing methods merely regard MLLMs as feature extractors or caption generators, leaving their capabilities in person ReID tasks largely unexplored. To bridge this gap, we introduce a novel benchmark for \underline{\textbf{V}}ersatile \underline{\textbf{P}}erson \underline{\textbf{Re}}-\underline{\textbf{ID}}entification, termed VP-ReID. The benchmark includes 257,310 multi-modal queries and gallery images, covering ten diverse person ReID tasks. In addition, we propose two task-oriented evaluation schemes for MLLM-based person ReID. Extensive experiments demonstrate the impressive versatility, effectiveness, and interpretability of MLLMs in various person ReID tasks. Nevertheless, they also have limitations in handling a few modalities, particularly thermal and infrared data. We hope that VP-ReID can facilitate the community in developing more robust and generalizable cross-modal foundation models for person ReID.




LLM-based Agents Suffer from Hallucinations: A Survey of Taxonomy, Methods, and Directions

Lin, Xixun, Ning, Yucheng, Zhang, Jingwen, Dong, Yan, Liu, Yilong, Wu, Yongxuan, Qi, Xiaohua, Sun, Nan, Shang, Yanmin, Wang, Kun, Cao, Pengfei, Wang, Qingyue, Zou, Lixin, Chen, Xu, Zhou, Chuan, Wu, Jia, Zhang, Peng, Wen, Qingsong, Pan, Shirui, Wang, Bin, Cao, Yanan, Chen, Kai, Hu, Songlin, Guo, Li

arXiv.org Artificial Intelligence

Abstract--Driven by the rapid advancements of Large Language Models (LLMs), LLM-based agents have emerged as powerful intelligent systems capable of human-like cognition, reasoning, and interaction. These agents are increasingly being deployed across diverse real-world applications, including student education, scientific research, and financial analysis. However, despite their remarkable potential, LLM-based agents remain vulnerable to hallucination issues, which can result in erroneous task execution and undermine the reliability of the overall system design. Addressing this critical challenge requires a deep understanding and a systematic consolidation of recent advances on LLM-based agents. T o this end, we present the first comprehensive survey of hallucinations in LLM-based agents. By carefully analyzing the complete workflow of agents, we propose a new taxonomy that identifies different types of agent hallucinations occurring at different stages. Furthermore, we conduct an in-depth examination of eighteen triggering causes underlying the emergence of agent hallucinations. Through a detailed review of a large number of existing studies, we summarize approaches for hallucination mitigation and detection, and highlight promising directions for future research. We hope this survey will inspire further efforts toward addressing hallucinations in LLM-based agents, ultimately contributing to the development of more robust and reliable agent systems. Cao, K. Chen, S. Hu, and L. Guo are with Institute of Information Engineering, Chinese Academy of Sciences, School of Cyber Security, University of Chinese Academy of Sciences, Beijing, China. K. Wang is with Nanyang Technological University, Singapore. Cao is with Institute of Automation, Chinese Academy of Sciences, Beijing, China. Q. Wang is with Hong Kong University of Science and Technology, Hong Kong, China. L. Zou is with School of Cyber Science and Engineering, Wuhan University, Wuhan, China. X. Chen is with Gaoling School of Artificial Intelligence, Renmin University of China, Beijing, China. C. Zhou is with Academy of Mathematics and Systems Science, Chinese Academy of Sciences, Beijing, China. J. Wu is with School of Computing, Faculty of Science and Engineering, Macquarie University, Sydney, Australia. Zhang is with the Cyberspace Institute of Advanced Technology, Guangzhou University, Guangzhou, China. Q. Wen is with Squirrel Ai Learning, Bellevue, USA. S. Pan is with School of Information and Communication Technology, Griffith University, Gold Coast, Australia. B. Wang is with Xiaomi Company, Beijing, China.